Goto

Collaborating Authors

 principal minor




A simple estimator of the correlation kernel matrix of a determinantal point process

Gouriéroux, Christian, Lu, Yang

arXiv.org Machine Learning

Determinantal Point Process (DPP) is a flexible family of distributions for random sets defined on the finite state space { 1, ...,d }, or equivalently for multivariate binary variables. This family is parameterized by either the L-ensemble kernel Σ, which is symmetric positive definite (SPD), or the correlation kernel matrix K, which is SPD, with eigenvalues lying strictly between 0 and 1. The literature has considered the maximum likelihood estimation (MLE) of Σ and K or its algorithmic analogues (Affandi et al., 2014; Brunel et al., 2017a,b), but it has since been shown that i) the likelihood function has at least 2


Efficient Tomography of Non-Interacting Fermion States

Aaronson, Scott, Grewal, Sabee

arXiv.org Artificial Intelligence

There are two types of particles in the universe: bosons and fermions. Bosons include force carriers, such as photons and gluons, and fermions include matter particles like quarks and electrons. Each particle can be in a certain mode (e.g., a position or state). For a system of n particles, a configuration of the system is described by specifying how many particles are in each of m modes. Bosons are particles where multiple occupancy of a mode is allowed, whereas fermions are particles where multiple occupancy is forbidden; that is, two or more fermions cannot occupy the same mode at once (this is the Pauli exclusion principle).


Linear Regression

#artificialintelligence

Regression is an Algorithm of the Supervised Learning model. When the output or the dependent feature is continuous and labeled then, we apply the Regression Algorithm. Regression is used to find the relation or equation between the Independent variables and the output variable. E.g., given below, we have variable x₁, x₂, ….,xₙ, which contribute towards the output of variable y. We have to find a relation between x variables and dependent variable y.


WISDoM: a framework for the Analysis of Wishart distributed matrices

Mengucci, Carlo, Remondini, Daniel, Giampieri, Enrico

arXiv.org Machine Learning

APPENDIX A. Visualizing the Wishart Distribution The Wishart distribution is a generalization to multiple dimensions of the chi-squared distribution, or in the case of non-integer degrees of freedom, of the gamma distribution. We show in fig.5 that for a 1-dimensional and equal to 1 Σ scale matrix, the Wishart distribution W 1( n, 1) is equivalent to the χ 2 ( n) distribution. Figure 5: Monodimensional Wishart Distribution and χ 2 (n) distribution comparison Save for this simple case, being the Wishart a distribution over matrices, it is a generally hard task to visualize it as a density function. Samples can be however drawn from it and the eigenvectors and eigenvalues of the resulting sampled matrix can be exploited to define an ellipse. An example of this technique is shown in fig.6.


Learning Signed Determinantal Point Processes through the Principal Minor Assignment Problem

Brunel, Victor-Emmanuel

Neural Information Processing Systems

Symmetric determinantal point processes (DPP) are a class of probabilistic models that encode the random selection of items that have a repulsive behavior. They have attracted a lot of attention in machine learning, where returning diverse sets of items is sought for. Sampling and learning these symmetric DPP's is pretty well understood. In this work, we consider a new class of DPP's, which we call signed DPP's, where we break the symmetry and allow attractive behaviors. We set the ground for learning signed DPP's through a method of moments, by solving the so called principal assignment problem for a class of matrices $K$ that satisfy $K_{i,j}=\pm K_{j,i}$, $i\neq j$, in polynomial time.


Learning Signed Determinantal Point Processes through the Principal Minor Assignment Problem

Brunel, Victor-Emmanuel

Neural Information Processing Systems

Symmetric determinantal point processes (DPP) are a class of probabilistic models that encode the random selection of items that have a repulsive behavior. They have attracted a lot of attention in machine learning, where returning diverse sets of items is sought for. Sampling and learning these symmetric DPP's is pretty well understood. In this work, we consider a new class of DPP's, which we call signed DPP's, where we break the symmetry and allow attractive behaviors. We set the ground for learning signed DPP's through a method of moments, by solving the so called principal assignment problem for a class of matrices $K$ that satisfy $K_{i,j}=\pm K_{j,i}$, $i\neq j$, in polynomial time.


A sufficient condition on monotonic increase of the number of nonzero entry in the optimizer of L1 norm penalized least-square problem

Duan, J., Soussen, Charles, Brie, David, Idier, Jerome, Wang, Y. -P.

arXiv.org Machine Learning

The $\ell$-1 norm based optimization is widely used in signal processing, especially in recent compressed sensing theory. This paper studies the solution path of the $\ell$-1 norm penalized least-square problem, whose constrained form is known as Least Absolute Shrinkage and Selection Operator (LASSO). A solution path is the set of all the optimizers with respect to the evolution of the hyperparameter (Lagrange multiplier). The study of the solution path is of great significance in viewing and understanding the profile of the tradeoff between the approximation and regularization terms. If the solution path of a given problem is known, it can help us to find the optimal hyperparameter under a given criterion such as the Akaike Information Criterion. In this paper we present a sufficient condition on $\ell$-1 norm penalized least-square problem. Under this sufficient condition, the number of nonzero entries in the optimizer or solution vector increases monotonically when the hyperparameter decreases. We also generalize the result to the often used total variation case, where the $\ell$-1 norm is taken over the first order derivative of the solution vector. We prove that the proposed condition has intrinsic connections with the condition given by Donoho, et al \cite{Donoho08} and the positive cone condition by Efron {\it el al} \cite{Efron04}. However, the proposed condition does not need to assume the sparsity level of the signal as required by Donoho et al's condition, and is easier to verify than Efron, et al's positive cone condition when being used for practical applications.